4 research outputs found
DataLev: Mid-air Data Physicalisation Using Acoustic Levitation
Data physicalisation is a technique that encodes data through the
geometric and material properties of an artefact, allowing users
to engage with data in a more immersive and multi-sensory way.
However, current methods of data physicalisation are limited in
terms of their reconfgurability and the types of materials that can
be used. Acoustophoresis—a method of suspending and manipulating materials using sound waves—ofers a promising solution
to these challenges. In this paper, we present DataLev, a design
space and platform for creating reconfgurable, multimodal data
physicalisations with enriched materiality using acoustophoresis.
We demonstrate the capabilities of DataLev through eight examples and evaluate its performance in terms of reconfgurability and
materiality. Our work ofers a new approach to data physicalisation, enabling designers to create more dynamic, engaging, and
expressive artefacts
Estimating pilots’ cognitive load from ocular parameters through simulation and in-flight studies
Eye tracking is the process of measuring either the point of gaze (where one is looking) or the motion of an eye relative to the head. This paper investigated use of eye gaze trackers in military aviation environment to automatically estimate pilot’s cognitive load from ocular parameters. We used a fixed base variable stability flight simulator with longitudinal tracking task and collected data from 14 military pilots. In a second study, we undertook three test flights with a BAES Hawk Trainer aircraft doing air to ground attack training missions and constant G level turn maneuvers up to +5G. Our study found that ocular parameters like rate of fixation is significantly different in different flying conditions and significantly correlate with altitude gradient during air to ground dive training task, normal load factor (G) of the aircraft during constant G level turn maneuvers and pilot’s control inceptor and tracking error in simulation tasks. Results from our studies can be used for real time estimation of pilots’ cognitive load, providing suitable warnings and alerts to the pilot in cockpit and training of military pilots on cognitive load management during operational missions
Eye Gaze Controlled Projected Display in Automotive and Military Aviation Environments
This paper presents an eye gaze controlled projected display that can be used in aviation and automotive environment as a head up display. We have presented details of the hardware and software used in developing the display and an algorithm to improve performance of point and selection tasks in eye gaze controlled graphical user interface. The algorithm does not require changing layout of an interface; it rather puts a set of hotspots on clickable targets using a Simulated Annealing algorithm. Four user studies involving driving and flight simulators have found that the proposed projected display can improve driving and flying performance and significantly reduce pointing and selection times for secondary mission control tasks compared to existing interaction systems